surgical instrument
High-Precision Surgical Robotic System for Intraocular Procedures
Lai, Yu-Ting, Rosen, Jacob, Foroutani, Yasamin, Ma, Ji, Wu, Wen-Cheng, Hubschman, Jean-Pierre, Tsao, Tsu-Chin
Abstract--Despite the extensive demonstration of robotic systems for both cataract and vitreoretinal procedures, existing technologies or mechanisms still possess insufficient accuracy, precision, and degrees of freedom for instrument manipulation or potentially automated tool exchange during surgical procedures. A new robotic system that focuses on improving tooltip accuracy, tracking performance, and smooth instrument exchange mechanism is therefore designed and manufactured. Its tooltip accuracy, precision, and mechanical capability of maintaining small incision through remote center of motion were externally evaluated using an optical coherence tomography (OCT) system. Through robot calibration and precise coordinate registration, the accuracy of tooltip positioning was measured to be 0.053 0.031 mm, and the overall performance was demonstrated on an OCT - guided automated cataract lens extraction procedure with deep learning-based pre-operative anatomical modeling and real-time supervision. Surgical robots have demonstrated success in improving the safety and efficiency of surgical procedures in recent decades [1]. Intraocular procedures often require high precision and delicate manipulation to reduce surgical complications, but some procedures remain infeasible to humans due to physiological limitations [2], left alone retinal vein cannulation that is too risky to be performed by human surgeons [3]. Even experienced surgeons include an average hand tremor of 200 to 350 µm relative to vein diameters of 120 to 200 µm, as well as the inability to perceive the small manipulating forces associated with piercing veins [4], [5]. As a result, surgical robots become good candidates for this type of sophisticated tissue manipulation.
- North America > United States > California > Los Angeles County > Los Angeles (0.28)
- North America > United States > Massachusetts > Middlesex County > Natick (0.04)
- Health & Medicine > Therapeutic Area > Ophthalmology/Optometry (1.00)
- Health & Medicine > Surgery (1.00)
Real-Time Surgical Instrument Defect Detection via Non-Destructive Testing
Ain, Qurrat Ul, Jilani, Atif Aftab Ahmed, Shafqat, Zunaira, Butt, Nigar Azhar
Defective surgical instruments pose serious risks to sterility, mechanical integrity, and patient safety, increasing the likelihood of surgical complications. However, quality control in surgical instrument manufacturing often relies on manual inspection, which is prone to human error and inconsistency. This study introduces SurgScan, an AI-powered defect detection framework for surgical instruments. Using YOLOv8, SurgScan classifies defects in real-time, ensuring high accuracy and industrial scalability. The model is trained on a high-resolution dataset of 102,876 images, covering 11 instrument types and five major defect categories. Extensive evaluation against state-of-the-art CNN architectures confirms that SurgScan achieves the highest accuracy (99.3%) with real-time inference speeds of 4.2-5.8 ms per image, making it suitable for industrial deployment. Statistical analysis demonstrates that contrast-enhanced preprocessing significantly improves defect detection, addressing key limitations in visual inspection. SurgScan provides a scalable, cost-effective AI solution for automated quality control, reducing reliance on manual inspection while ensuring compliance with ISO 13485 and FDA standards, paving the way for enhanced defect detection in medical manufacturing.
- North America > United States (0.34)
- North America > Canada > Quebec > Capitale-Nationale Region > Québec (0.04)
- North America > Canada > Quebec > Capitale-Nationale Region > Quebec City (0.04)
- (2 more...)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (1.00)
- Health & Medicine > Surgery (1.00)
- Health & Medicine > Health Care Technology (1.00)
- Government > Regional Government > North America Government > United States Government > FDA (0.34)
Surgical tools could get a bug-inspired upgrade
Amazon Prime Day is live. See the best deals HERE. Female sawflies can cut into a plant's tissue without destroying the rest of the plant. Breakthroughs, discoveries, and DIY tips sent every weekday. If you're one of those people who serve as a veritable mosquito buffet in the summer, it might seem like insects just bite through skin indiscriminately.
- Oceania > New Zealand (0.05)
- North America > United States > Texas (0.05)
- Europe > United Kingdom > Scotland (0.05)
- Health & Medicine > Surgery (0.68)
- Health & Medicine > Health Care Technology (0.54)
NeeCo: Image Synthesis of Novel Instrument States Based on Dynamic and Deformable 3D Gaussian Reconstruction
Zeng, Tianle, Hu, Junlei, Galindo, Gerardo Loza, Ali, Sharib, Sarikaya, Duygu, Valdastri, Pietro, Jones, Dominic
This work has been submitted to the IEEE for possible publication. Abstract --Computer vision-based technologies significantly enhance surgical automation by advancing tool tracking, detection, and localization. However, Current data-driven approaches are data-voracious, requiring large, high-quality labeled image datasets, which limits their application in surgical data science. Our Work introduces a novel dynamic Gaussian Splatting technique to address the data scarcity in surgical image datasets. We propose a dynamic Gaussian model to represent dynamic surgical scenes, enabling the rendering of surgical instruments from unseen viewpoints and deformations with real tissue backgrounds. We utilize a dynamic training adjustment strategy to address challenges posed by poorly calibrated camera poses from real-world scenarios. Additionally, we propose a method based on dynamic Gaussians for automatically generating annotations for our synthetic data. For evaluation, we constructed a new dataset featuring seven scenes with 14,000 frames of tool and camera motion and tool jaw articulation, with a background of an ex-vivo porcine model. Using this dataset, we synthetically replicate the scene deformation from the ground truth data, allowing direct comparisons of synthetic image quality. Experimental results illustrate that our method generates photo-realistic labeled image datasets with the highest values in Peak-Signal-to-Noise Ratio (29.87). We further evaluate the performance of medical-specific neural networks trained on real and synthetic images using an unseen real-world image dataset. Our results show that the performance of models trained on synthetic images generated by the proposed method outperforms those trained with state-of-the-art standard data augmentation by 10%, leading to an overall improvement in model performances by nearly 15%. OMPUTER vision plays a crucial role in advancing intelligent surgical navigation, planning, and automation [1].
- Europe > United Kingdom > England > West Yorkshire > Leeds (0.04)
- Asia > China (0.04)
- North America > Canada (0.04)
- Europe > Netherlands > North Holland > Amsterdam (0.04)
- Health & Medicine > Surgery (1.00)
- Health & Medicine > Health Care Technology (0.91)
- Health & Medicine > Diagnostic Medicine > Imaging (0.69)
Can DeepSeek Reason Like a Surgeon? An Empirical Evaluation for Vision-Language Understanding in Robotic-Assisted Surgery
Ma, Boyi, Zhao, Yanguang, Wang, Jie, Wang, Guankun, Yuan, Kun, Chen, Tong, Bai, Long, Ren, Hongliang
The DeepSeek models have shown exceptional performance in general scene understanding, question-answering (QA), and text generation tasks, owing to their efficient training paradigm and strong reasoning capabilities. In this study, we investigate the dialogue capabilities of the DeepSeek model in robotic surgery scenarios, focusing on tasks such as Single Phrase QA, Visual QA, and Detailed Description. The Single Phrase QA tasks further include sub-tasks such as surgical instrument recognition, action understanding, and spatial position analysis. We conduct extensive evaluations using publicly available datasets, including EndoVis18 and CholecT50, along with their corresponding dialogue data. Our empirical study shows that, compared to existing general-purpose multimodal large language models, DeepSeek-VL2 performs better on complex understanding tasks in surgical scenes. Additionally, although DeepSeek-V3 is purely a language model, we find that when image tokens are directly inputted, the model demonstrates better performance on single-sentence QA tasks. However, overall, the DeepSeek models still fall short of meeting the clinical requirements for understanding surgical scenes. Under general prompts, DeepSeek models lack the ability to effectively analyze global surgical concepts and fail to provide detailed insights into surgical scenarios. Based on our observations, we argue that the DeepSeek models are not ready for vision-language tasks in surgical contexts without fine-tuning on surgery-specific datasets.
- Europe > France > Grand Est > Bas-Rhin > Strasbourg (0.04)
- Asia > China > Hong Kong (0.04)
- Europe > Netherlands > North Holland > Amsterdam (0.04)
- Health & Medicine > Surgery (1.00)
- Health & Medicine > Health Care Technology (1.00)
- Health & Medicine > Diagnostic Medicine > Imaging (0.47)
An analysis of higher-order kinematics formalisms for an innovative surgical parallel robot
Vaida, Calin, Birlescu, Iosif, Gherman, Bogdan, Condurache, Daniel, Chablat, Damien, Pisla, Doina
The paper presents a novel modular hybrid parallel robot for pancreatic surgery and its higher-order kinematics derived based on various formalisms. The classical vector, homogeneous transformation matrices and dual quaternion approaches are studied for the kinematic functions using both classical differentiation and multidual algebra. The algorithms for inverse kinematics for all three studied formalisms are presented for both differentiation and multidual algebra approaches. Furthermore, these algorithms are compared based on numerical stability, execution times and number and type of mathematical functions and operators contained in each algorithm. A statistical analysis shows that there is significant improvement in execution time for the algorithms implemented using multidual algebra, while the numerical stability is appropriate for all algorithms derived based on differentiation and multidual algebra. While the implementation of the kinematic algorithms using multidual algebra shows positive results when benchmarked on a standard PC, further work is required to evaluate the multidual algorithms on hardware/software used for the modular parallel robot command and control.
- Europe > France > Pays de la Loire > Loire-Atlantique > Nantes (0.04)
- Europe > Romania > Nord-Vest Development Region > Cluj County > Cluj-Napoca (0.04)
- North America > United States > Missouri > St. Louis County > St. Louis (0.04)
- (4 more...)
- Health & Medicine > Therapeutic Area > Oncology (1.00)
- Health & Medicine > Surgery (1.00)
- Health & Medicine > Health Care Technology (0.94)
Instrument-Splatting: Controllable Photorealistic Reconstruction of Surgical Instruments Using Gaussian Splatting
Yang, Shuojue, Wu, Zijian, Hong, Mingxuan, Li, Qian, Shen, Daiyun, Salcudean, Septimiu E., Jin, Yueming
Real2Sim is becoming increasingly important with the rapid development of surgical artificial intelligence (AI) and autonomy. In this work, we propose a novel Real2Sim methodology, Instrument-Splatting, that leverages 3D Gaussian Splatting to provide fully controllable 3D reconstruction of surgical instruments from monocular surgical videos. To maintain both high visual fidelity and manipulability, we introduce a geometry pre-training to bind Gaussian point clouds on part mesh with accurate geometric priors and define a forward kinematics to control the Gaussians as flexible as real instruments. Afterward, to handle unposed videos, we design a novel instrument pose tracking method leveraging semantics-embedded Gaussians to robustly refine per-frame instrument poses and joint states in a render-and-compare manner, which allows our instrument Gaussian to accurately learn textures and reach photorealistic rendering. We validated our method on 2 publicly released surgical videos and 4 videos collected on ex vivo tissues and green screens. Quantitative and qualitative evaluations demonstrate the effectiveness and superiority of the proposed method.
- Health & Medicine > Surgery (0.60)
- Health & Medicine > Health Care Technology (0.60)
SurgPose: a Dataset for Articulated Robotic Surgical Tool Pose Estimation and Tracking
Wu, Zijian, Schmidt, Adam, Moore, Randy, Zhou, Haoying, Banks, Alexandre, Kazanzides, Peter, Salcudean, Septimiu E.
Accurate and efficient surgical robotic tool pose estimation is of fundamental significance to downstream applications such as augmented reality (AR) in surgical training and learning-based autonomous manipulation. While significant advancements have been made in pose estimation for humans and animals, it is still a challenge in surgical robotics due to the scarcity of published data. The relatively large absolute error of the da Vinci end effector kinematics and arduous calibration procedure make calibrated kinematics data collection expensive. Driven by this limitation, we collected a dataset, dubbed SurgPose, providing instance-aware semantic keypoints and skeletons for visual surgical tool pose estimation and tracking. By marking keypoints using ultraviolet (UV) reactive paint, which is invisible under white light and fluorescent under UV light, we execute the same trajectory under different lighting conditions to collect raw videos and keypoint annotations, respectively. The SurgPose dataset consists of approximately 120k surgical instrument instances (80k for training and 40k for validation) of 6 categories. Each instrument instance is labeled with 7 semantic keypoints. Since the videos are collected in stereo pairs, the 2D pose can be lifted to 3D based on stereo-matching depth. In addition to releasing the dataset, we test a few baseline approaches to surgical instrument tracking to demonstrate the utility of SurgPose. More details can be found at surgpose.github.io.
- Europe > Switzerland > Zürich > Zürich (0.14)
- North America > United States > California > Santa Clara County > Sunnyvale (0.04)
- North America > United States > California > Riverside County > Murrieta (0.04)
- (6 more...)
- Health & Medicine > Surgery (1.00)
- Health & Medicine > Health Care Technology (1.00)
- Health & Medicine > Diagnostic Medicine > Imaging (0.47)
Anatomy Might Be All You Need: Forecasting What to Do During Surgery
Sarwin, Gary, Carretta, Alessandro, Staartjes, Victor, Zoli, Matteo, Mazzatenta, Diego, Regli, Luca, Serra, Carlo, Konukoglu, Ender
Surgical guidance can be delivered in various ways. In neurosurgery, spatial guidance and orientation are predominantly achieved through neuronavigation systems that reference pre-operative MRI scans. Recently, there has been growing interest in providing live guidance by analyzing video feeds from tools such as endoscopes. Existing approaches, including anatomy detection, orientation feedback, phase recognition, and visual question-answering, primarily focus on aiding surgeons in assessing the current surgical scene. This work aims to provide guidance on a finer scale, aiming to provide guidance by forecasting the trajectory of the surgical instrument, essentially addressing the question of what to do next. To address this task, we propose a model that not only leverages the historical locations of surgical instruments but also integrates anatomical features. Importantly, our work does not rely on explicit ground truth labels for instrument trajectories. Instead, the ground truth is generated by a detection model trained to detect both anatomical structures and instruments within surgical videos of a comprehensive dataset containing pituitary surgery videos. By analyzing the interaction between anatomy and instrument movements in these videos and forecasting future instrument movements, we show that anatomical features are a valuable asset in addressing this challenging task. To the best of our knowledge, this work is the first attempt to address this task for manually operated surgeries.
- Europe > Switzerland > Zürich > Zürich (0.15)
- Europe > Italy > Emilia-Romagna > Metropolitan City of Bologna > Bologna (0.05)
- Health & Medicine > Surgery (1.00)
- Health & Medicine > Health Care Technology (1.00)
- Health & Medicine > Diagnostic Medicine > Imaging (0.54)
- Health & Medicine > Therapeutic Area > Neurology (0.49)
Virtual-Work Based Shape-Force Sensing for Continuum Instruments with Tension-Feedback Actuation
Zhang, Guoqing, Chen, Zihan, Wang, Long
Continuum instruments are integral to robot-assisted minimally invasive surgery (MIS), with tendon-driven mechanisms being the most common. Real-time tension feedback is crucial for precise articulation but remains a challenge in compact actuation unit designs. Additionally, accurate shape and external force sensing of continuum instruments are essential for advanced control and manipulation. This paper presents a compact and modular actuation unit that integrates a torque cell directly into the pulley module to provide real-time tension feedback. Building on this unit, we propose a novel shape-force sensing framework that incorporates polynomial curvature kinematics to accurately model non-constant curvature. The framework combines pose sensor measurements at the instrument tip and actuation tension feedback at the developed actuation unit. Experimental results demonstrate the improved performance of the proposed shape-force sensing framework in terms of shape reconstruction accuracy and force estimation reliability compared to conventional constant-curvature methods.
- North America > United States (0.04)
- Asia > China > Hong Kong (0.04)